我们提出了一种隐含的可能性方法,可以通过分散目录数据量化宇宙学信息,并作为图形组装。为此,我们使用模拟暗物质光环目录探索宇宙学的推断。我们采用最大化神经网络(IMNN)的信息来量化Fisher信息提取,这是图表的函数。我们a)在无噪声限制下,模块图结构对基础宇宙学具有高度敏感性,b)表明,通过比较传统统计,网络自动结合质量和聚类信息,c)证明图形神经网络仍然可以提取信息。当目录受到嘈杂的调查削减时,d)说明了如何将非线性IMNN摘要用作贝叶斯隐性可能性推断的渐近最佳压缩统计。我们在两点相关功能上,我们将$ \ omega_m,\ sigma_8 $参数约束降低了42倍,并证明网络自动组合质量和聚类信息,将关节$ \ omega_m,\ sigma_8 $参数约束减少42倍。 。这项工作利用了JAX中的图形数据的新IMNN实现,该实现可以利用数值或自动差异性。我们还显示,IMNNS成功地压缩了远离拟合网络的基准模型的模拟,这表明基于目录的分析中$ n $ point统计的有希望的替代方法。
translated by 谷歌翻译
Most benchmarks for studying surgical interventions focus on a specific challenge instead of leveraging the intrinsic complementarity among different tasks. In this work, we present a new experimental framework towards holistic surgical scene understanding. First, we introduce the Phase, Step, Instrument, and Atomic Visual Action recognition (PSI-AVA) Dataset. PSI-AVA includes annotations for both long-term (Phase and Step recognition) and short-term reasoning (Instrument detection and novel Atomic Action recognition) in robot-assisted radical prostatectomy videos. Second, we present Transformers for Action, Phase, Instrument, and steps Recognition (TAPIR) as a strong baseline for surgical scene understanding. TAPIR leverages our dataset's multi-level annotations as it benefits from the learned representation on the instrument detection task to improve its classification capacity. Our experimental results in both PSI-AVA and other publicly available databases demonstrate the adequacy of our framework to spur future research on holistic surgical scene understanding.
translated by 谷歌翻译
Video provides us with the spatio-temporal consistency needed for visual learning. Recent approaches have utilized this signal to learn correspondence estimation from close-by frame pairs. However, by only relying on close-by frame pairs, those approaches miss out on the richer long-range consistency between distant overlapping frames. To address this, we propose a self-supervised approach for correspondence estimation that learns from multiview consistency in short RGB-D video sequences. Our approach combines pairwise correspondence estimation and registration with a novel SE(3) transformation synchronization algorithm. Our key insight is that self-supervised multiview registration allows us to obtain correspondences over longer time frames; increasing both the diversity and difficulty of sampled pairs. We evaluate our approach on indoor scenes for correspondence estimation and RGB-D pointcloud registration and find that we perform on-par with supervised approaches.
translated by 谷歌翻译
A quantitative assessment of the global importance of an agent in a team is as valuable as gold for strategists, decision-makers, and sports coaches. Yet, retrieving this information is not trivial since in a cooperative task it is hard to isolate the performance of an individual from the one of the whole team. Moreover, it is not always clear the relationship between the role of an agent and his personal attributes. In this work we conceive an application of the Shapley analysis for studying the contribution of both agent policies and attributes, putting them on equal footing. Since the computational complexity is NP-hard and scales exponentially with the number of participants in a transferable utility coalitional game, we resort to exploiting a-priori knowledge about the rules of the game to constrain the relations between the participants over a graph. We hence propose a method to determine a Hierarchical Knowledge Graph of agents' policies and features in a Multi-Agent System. Assuming a simulator of the system is available, the graph structure allows to exploit dynamic programming to assess the importances in a much faster way. We test the proposed approach in a proof-of-case environment deploying both hardcoded policies and policies obtained via Deep Reinforcement Learning. The proposed paradigm is less computationally demanding than trivially computing the Shapley values and provides great insight not only into the importance of an agent in a team but also into the attributes needed to deploy the policy at its best.
translated by 谷歌翻译
Transfer learning on edge is challenging due to on-device limited resources. Existing work addresses this issue by training a subset of parameters or adding model patches. Developed with inference in mind, Inverted Residual Blocks (IRBs) split a convolutional layer into depthwise and pointwise convolutions, leading to more stacking layers, e.g., convolution, normalization, and activation layers. Though they are efficient for inference, IRBs require that additional activation maps are stored in memory for training weights for convolution layers and scales for normalization layers. As a result, their high memory cost prohibits training IRBs on resource-limited edge devices, and making them unsuitable in the context of transfer learning. To address this issue, we present MobileTL, a memory and computationally efficient on-device transfer learning method for models built with IRBs. MobileTL trains the shifts for internal normalization layers to avoid storing activation maps for the backward pass. Also, MobileTL approximates the backward computation of the activation layer (e.g., Hard-Swish and ReLU6) as a signed function which enables storing a binary mask instead of activation maps for the backward pass. MobileTL fine-tunes a few top blocks (close to output) rather than propagating the gradient through the whole network to reduce the computation cost. Our method reduces memory usage by 46% and 53% for MobileNetV2 and V3 IRBs, respectively. For MobileNetV3, we observe a 36% reduction in floating-point operations (FLOPs) when fine-tuning 5 blocks, while only incurring a 0.6% accuracy reduction on CIFAR10. Extensive experiments on multiple datasets demonstrate that our method is Pareto-optimal (best accuracy under given hardware constraints) compared to prior work in transfer learning for edge devices.
translated by 谷歌翻译
Obtaining photorealistic reconstructions of objects from sparse views is inherently ambiguous and can only be achieved by learning suitable reconstruction priors. Earlier works on sparse rigid object reconstruction successfully learned such priors from large datasets such as CO3D. In this paper, we extend this approach to dynamic objects. We use cats and dogs as a representative example and introduce Common Pets in 3D (CoP3D), a collection of crowd-sourced videos showing around 4,200 distinct pets. CoP3D is one of the first large-scale datasets for benchmarking non-rigid 3D reconstruction "in the wild". We also propose Tracker-NeRF, a method for learning 4D reconstruction from our dataset. At test time, given a small number of video frames of an unseen object, Tracker-NeRF predicts the trajectories of its 3D points and generates new views, interpolating viewpoint and time. Results on CoP3D reveal significantly better non-rigid new-view synthesis performance than existing baselines.
translated by 谷歌翻译
尽管深度神经网络(DNNS)具有很大的概括和预测能力,但它们的功能不允许对其行为进行详细的解释。不透明的深度学习模型越来越多地用于在关键环境中做出重要的预测,而危险在于,它们做出和使用不能合理或合法化的预测。已经出现了几种可解释的人工智能(XAI)方法,这些方法与机器学习模型分开了,但对模型的实际功能和鲁棒性具有忠诚的缺点。结果,就具有解释能力的深度学习模型的重要性达成了广泛的协议,因此他们自己可以为为什么做出特定的预测提供答案。首先,我们通过形式化解释是什么是缺乏XAI的普遍标准的问题。我们还引入了一组公理和定义,以从数学角度阐明XAI。最后,我们提出了Greybox XAI,该框架由于使用了符号知识库(KB)而构成DNN和透明模型。我们从数据集中提取KB,并使用它来训练透明模型(即逻辑回归)。在RGB图像上训练了编码器 - 编码器架构,以产生类似于透明模型使用的KB的输出。一旦两个模型被独立训练,它们就会在组合上使用以形成可解释的预测模型。我们展示了这种新体系结构在几个数据集中如何准确且可解释的。
translated by 谷歌翻译
社交网络数据评估的自动化是自然语言处理的经典挑战之一。在共同199年的大流行期间,关于了解健康命令的态度,公共信息中的采矿人们的立场变得至关重要。在本文中,作者提出了基于变压器体系结构的预测模型,以对Twitter文本中的前提进行分类。这项工作是作为2022年社交媒体挖掘(SMM4H)研讨会的一部分完成的。我们探索了现代变压器的分类器,以便构建管道有效地捕获推文语义。我们在Twitter数据集上的实验表明,在前提预测任务的情况下,罗伯塔(Roberta)优于其他变压器模型。该模型在ROC AUC值0.807方面实现了竞争性能,而F1得分为0.7648。
translated by 谷歌翻译
医学图像中的自动对象识别可以促进医学诊断和治疗。在本文中,我们自动对超声图像中的锁骨神经进行了分割,以帮助注入周围神经块。神经块通常用于手术后的疼痛治疗,其中使用超声指导在靶神经旁边注入局部麻醉药。这种治疗可以阻止疼痛信号向大脑的传播,这可以帮助提高手术中的恢复速率,并显着减少术后阿片类药物的需求。但是,超声引导的区域麻醉(UGRA)要求麻醉师在视觉上识别超声图像中的实际神经位置。鉴于超声图像中神经的无视觉效果以及它们与许多相邻组织的视觉相似性,这是一项复杂的任务。在这项研究中,我们使用了自动神经检测系统进行UGRA神经阻滞治疗。该系统可以使用深度学习技术识别神经在超声图像中的位置。我们开发了一个模型来捕获神经的特征,通过训练两个具有跳过连接的深神经网络:两种扩展的U-NET体系结构,有或没有扩张的卷积。该溶液可能会导致区域麻醉中靶向神经的封锁。
translated by 谷歌翻译
在治疗方案中应用的机器人,例如在自闭症谱系障碍的个体治疗中,有时被用于模仿学习活动,其中一个人需要由机器人重复运动。为了简化合并机器人可以执行的新运动的任务,希望机器人能够通过观察人类(例如治疗师)的示威来学习动作。在本文中,我们研究了一种从人类的骨骼观察中获取动作的方法,该方法是由以机器人为中心的RGB-D摄像头收集的。给定一系列观察到各种关节,在通过PID位置控制器执行之前,将关节位置映射以匹配机器人的配置。我们通过使用Qtrobot进行一项研究来评估该方法,尤其是繁殖误差,其中机器人从多个参与者中获取了不同的上身舞蹈动作。结果表明该方法的总体可行性,但也表明繁殖质量受骨架观测中噪声的影响。
translated by 谷歌翻译